11 research outputs found

    #4 CRAWLING VON TEXTDATEN MIT DDC, LCC BEZUG ZUR GENERIERUNG EINER TRAININGSDATENMENGE FÜR DIE TEXTKLASSIFIKATION: Praktikumsbericht Textmining – Wissensrohstoff Text

    Get PDF
    Ziel unseres Berichts ist die Evaluation der Datenverfügbarkeit und das Erstellen eines Datensatzes, der später zum maschinellen Lernen von Bibliotheksklassifikationen genutzt werden könnte. Als Basis für die Textdaten werden wir Wikidata-Einträge nutzen, da diese teilweise bereits mit solchen Klassifikationen versehen und direkt mit dem zugehörigen Wikipedia-Artikel verknüpft sind

    The Freedom Collection 2017–2021: Part 1, The Composition of the Freedom Collection and UNCL’s Downloads by Member and Subject

    Get PDF
    With data provided by a colleague, the author looked at the University of Nebraska Consortium of Libraries (UNCL) downloading activity of Elsevier\u27s Freedom Collection for the 2017-2021 interval. Members include the University of Nebraska at Kearney, University of Nebraska-Lincoln, University of Nebraska Medical Center, and University of Nebraska Omaha. The author looked at activity by member and by subject at the level of journals\u27 Library of Congress Classifications. This report was submitted in the fall of 2022 to the UNL Libraries Collections Strategies Committee (CSC) and to the members of the Collections Strategies and Open Scholarship (CSOS) department

    Modelling the Structure and Dynamics of Science Using Books

    Full text link
    Scientific research is a major driving force in a knowledge based economy. Income, health and wellbeing depend on scientific progress. The better we understand the inner workings of the scientific enterprise, the better we can prompt, manage, steer, and utilize scientific progress. Diverse indicators and approaches exist to evaluate and monitor research activities, from calculating the reputation of a researcher, institution, or country to analyzing and visualizing global brain circulation. However, there are very few predictive models of science that are used by key decision makers in academia, industry, or government interested to improve the quality and impact of scholarly efforts. We present a novel 'bibliographic bibliometric' analysis which we apply to a large collection of books relevant for the modelling of science. We explain the data collection together with the results of the data analyses and visualizations. In the final section we discuss how the analysis of books that describe different modelling approaches can inform the design of new models of science.Comment: data and large scale maps http://cns.iu.edu/2015-ModSci.html, Ginda, Michael, Andrea Scharnhorst, and Katy B\"orner. "Modelling Science". In Theories of Informetrics: A Festschrift in Honor of Blaise Cronin, edited by Sugimoto, Cassidy. Munich: De Gruyter Sau

    Bibliography versus Auto-Bibliography: Tackling the Transformation of Traditions in the Research Process

    Get PDF
    Ms. Babb reports on a study conducted to determine whether researchers will identify the same works recommended by scholarly bibliographies if their searching is limited to the confines of the library catalog and its subject headings. She explores how the auto-bibliography of the catalog compares to more traditionally compiled bibliographies, and what—if anything—is sacrificed when users rely upon auto-bibliography rather than scholarly bibliography

    Workset Creation for Scholarly Analysis and Data Capsules (WCSA+DC): Laying the foundations for secure computation with copyrighted data in the HathiTrust Research Center, Phase I

    Get PDF
    The primary objective of the WCSA+DC project is the seamless integration of the workset model and tools with the Data Capsule framework to provide non-consumptive research access HathiTrust’s massive corpus of data objects, securely and at scale, regardless of copyright status. That is, we plan to surmount the copyright wall on behalf of scholars and their students. Notwithstanding the substantial preliminary work that has been done on both the WCSA and DC fronts, they are both still best characterized as being in the prototyping stages. It is our intention to that this proposed Phase I of the project devote an intense two-year burst of effort to move the suite of WCSA and DC prototypes from the realm of proof-of-concept to that of a firmly integrated at-scale deployment. We plan to concentrate our requested resources on making sure our systems are as secure and robust at scale as possible. Phase I will engage four external research partners. Two of the external partners, Kevin Page (Oxford) and Annika Hinze (Waikato) were recipients of WCSA prototyping sub-awards. We are very glad to propose extending and refining aspects of their prototyping work in the context of WCSA+DC. Two other scholars, Ted Underwood (Illinois) and James Pustejovsky (Brandeis) will play critical roles in Phase I as active participants in the development and refinement of the tools and systems from their particular user-scholar perspectives: Underwood, Digital Humanities (DH); Pustejovsky, Computational Linguistics (CL). The four key outcomes and benefits of the WCSA+DC, Phase I project are: 1. The deployment of a new Workset Builder tool that enhances search and discovery across the entire HTDL by complementing traditional volume-level bibliographic metadata with new metadata derived from a variety of sources at various levels granularity. 2. The creation of Linked Open Data resources to help scholars find, select, integrate and disseminate a wider range of data as part of their scholarly analysis life-cycle. 3. A new Data Capsule framework that integrates worksets, runs at scale, and does both in a secure, non-consumptive, manner. 4. A set of exemplar pre-built Data Capsules that incorporate tools commonly used by both the DH and CL communities that scholars can then customize to their specific needs.Andrew W. Mellon Foundation, grant no. 41500672Ope

    Improving the Formatting Tools of CDS Invenio

    Get PDF
    CDS Invenio is the web-based integrated digital library system developed at CERN. It is a strategical tool that supports the archival and open dissemination of documents produced by CERN researchers. This paper reports on my Master’s thesis work done on BibFormat, a module in CDS Invenio, which formats documents metadata. The goal of this project was to implement a completely new formatting module for CDS Invenio. In this report a strong emphasis is put on the user-centered design of the new BibFormat. The bibliographic formatting process and its requirements are discussed. The task analysis and its resulting interaction model are detailed. The document also shows the implemented user interface of BibFormat and gives the results of the user evaluation of this interface. Finally the results of a small usability study of the formats included in CDS Invenio are discussed

    Theories of Informetrics and Scholarly Communication

    Get PDF
    Scientometrics have become an essential element in the practice and evaluation of science and research, including both the evaluation of individuals and national assessment exercises. Yet, researchers and practitioners in this field have lacked clear theories to guide their work. As early as 1981, then doctoral student Blaise Cronin published "The need for a theory of citing" —a call to arms for the fledgling scientometric community to produce foundational theories upon which the work of the field could be based. More than three decades later, the time has come to reach out the field again and ask how they have responded to this call. This book compiles the foundational theories that guide informetrics and scholarly communication research. It is a much needed compilation by leading scholars in the field that gathers together the theories that guide our understanding of authorship, citing, and impact

    Theories of Informetrics and Scholarly Communication

    Get PDF
    Scientometrics have become an essential element in the practice and evaluation of science and research, including both the evaluation of individuals and national assessment exercises. Yet, researchers and practitioners in this field have lacked clear theories to guide their work. As early as 1981, then doctoral student Blaise Cronin published The need for a theory of citing - a call to arms for the fledgling scientometric community to produce foundational theories upon which the work of the field could be based. More than three decades later, the time has come to reach out the field again and ask how they have responded to this call. This book compiles the foundational theories that guide informetrics and scholarly communication research. It is a much needed compilation by leading scholars in the field that gathers together the theories that guide our understanding of authorship, citing, and impact
    corecore